Evolutionary Ecology
○ Springer Science and Business Media LLC
Preprints posted in the last 7 days, ranked by how well they match Evolutionary Ecology's content profile, based on 14 papers previously published here. The average preprint has a 0.01% match score for this journal, so anything above that is already an above-average fit.
Teeling, C.; Vogeler, S.; Ellis, R. P.; Williams, E. A.
Show abstract
Settlement, the transition of a swimming planktonic larva to a crawling or sessile benthic juvenile, is a key process in the development of many marine invertebrates. Successful recruitment via larval settlement is critical for the development and maintenance of seafloor ecosystems. Microbial biofilms act as positive cues for larval settlement across diverse taxa, yet the behavioural processes preceding settlement are poorly understood. Here, we investigated age-dependent changes in settlement behaviour in the marine polychaete Platynereis dumerilii larvae in response to Grammatophora marina diatom biofilms. Settlement behaviours (crawling, crawling speed, and track straightness (tortuosity)) were quantified from recordings of larvae at five developmental stages (mid-trochophore to late-nectochaete) in the presence or absence of diatom biofilms, using image segmentation and spot-tracking software. As larvae developed, the proportion of individuals crawling (settlement) over the biofilm increased. Older larvae colonised biofilms more rapidly and showed greater discrimination between G. marina biofilms and non-biofilmed controls. The movement trajectory of older larvae also straightens compared to individuals swimming in the presence of biofilms, or behaviour witnessed in the absence of biofilms. The proportions and magnitudes of these behaviours may reflect changing prioritisation of sensory inputs from physical and chemical cues as larvae develop. Our findings suggest that behavioural traits that are associated with settlement are developmentally programmed in P. dumerilii. Understanding settlement behaviours in P. dumerilii expands on this species behavioural repertoire and sheds light on the evolutionary relationship between marine larvae and microalgal biofilms.
van Rooyen, N. T.; Prugnolle, F.; Rougeron, V.; Hofmeester, T. R.
Show abstract
Understanding how the fear of predation acts as a driver of spatial distribution is fundamental to animal behaviour research, yet this relationship is not wholly understood in primates such as baboons. Olive baboons (Papio anubis) have evolved a diverse range of antipredator strategies that reduce, but do not eliminate, predation risk from the large carnivores they encounter across their broad geographic range. This raises a critical question: does the need to access essential resources outweigh the risk of predation when determining habitat selection? We addressed this question by examining the relative influence of three environmental factors and relative predator abundance on olive baboon occupancy patterns and detection probability in Serengeti National Park, Tanzania. Using data from 225 camera traps deployed by the Snapshot Safari program, we fitted three separate Bayesian occupancy models, each incorporating the same three environmental covariates (terrain ruggedness index, distance to nearest river, and Normalized Difference Vegetation Index, NDVI), together with the relative abundance of one of three principal predators (lion, leopard, or spotted hyena). This approach allowed us to assess whether environmental covariates associated with baboon occupancy remained consistent across different predator contexts. Baboon occupancy strongly increased with terrain ruggedness in all three models and consistently decreased with a greater distance to rivers. Vegetation greenness (NDVI) showed a positive association with baboon occupancy, though credible intervals narrowly overlapped zero. NDVI also showed a strong positive relationship with baboon detection probability. Associations between predator relative abundance and baboon occupancy varied between models: the relative abundance of lions and spotted hyenas showed no strong association with baboon occupancy, whereas the relative abundance of leopards was strongly correlated with baboon occupancy, consistent with shared habitat preferences. Our findings demonstrate that, independent of predator presence, olive baboon spatial distribution in the Serengeti is primarily and consistently associated with resource-related environmental features. This study expands our knowledge on the ecological factors that influence primate occupancy by showing that, for a behaviourally flexible species with diverse antipredator strategies, access to essential resources can outweigh spatial avoidance of predators in a multi-predator landscape.
Zhan, Q.; Pascual, M.; He, Q.
Show abstract
Major surface antigens in many pathogens are encoded by rapidly diversifying multigene families, generating fitness variation through antigenic and functional differences. These variations align with the niche and absolute fitness axes of Modern Coexistence Theory (MCT). Yet, how such gene families evolve along these axes under competition for hosts and across transmission gradients remains poorly understood, as prior MCT studies have not explicitly accounted for evolutionary dynamics in high dimensions. We use a stochastic computational model of Plasmodium falciparum transmission to examine how transmission intensity and selection shape var multigene family evolution and composition within parasite genomes. Results show that selection alone cannot maintain the observed stable ratio of two gene groups within parasite genomes, indicating that group-based classifications do not clearly reflect transmission strategy or virulence. When a trade-off exists between diversification rates and absolute fitness, strong immune selection under high transmission favors fast-recombining genes while attenuating functional selection on R0-associated traits. In general, stronger immune selection increases the invasion probability of novel antigens and the niche differentiation among parasite genomes, while reducing the variance in gene-level transmissibility and expression duration, and therefore R0. This outcome, combining enhanced niche differentiation and reduced absolute fitness variation, departs from MCT predictions.
Bauman, A.; Owen, K.; Messing, S.; Macdonald, H.; Nettlefold, L.; Richards, J.; Vandelanotte, C.; Chen, I.-H.; Cullen, B.; van Buskirk, J.; van Itallie, A.; Coletta, G.; O'Halloran, P.; Randle, E.; Nicholson, M.; Staley, K.; McKay, H. A.
Show abstract
Military aviation training noise remains understudied despite its widespread impacts across urban, rural, and wilderness areas. The predominance of low-frequency noise and repetitive training can create pervasive noise pollution, yet past research often fails to capture the full range of health and quality-of-life effects. This study analyzed two complaint datasets related to Whidbey Island Naval Air Station noise: U.S. Navy records (2017-2020) and Quiet Skies Over San Juan County data (2021-2023). We analyzed and mapped sentiment intensity from noise complaints relative to modeled annual noise exposure, developed a typology to classify impacts, and modeled the environmental and operational factors influencing complaints. Findings revealed widespread negative sentiment and anger, often beyond the bounds of estimated noise contours, suggesting that annual cumulative noise models inadequately estimate community impacts. Complaints consistently highlighted sleep disturbance, hearing and health concerns, and compromised home environments due to shaking, vibration, and disruption of daily life. Residents also reported significant social, recreational, and work disruptions, along with feelings of fear, helplessness, and concern for children's well-being. The number of complaints were strongly associated with training schedules, with late-night sessions being the strongest predictor. A delayed response pattern suggests residents reach a frustration threshold before filing complaints. Overall, our findings demonstrate persistent negative sentiment and diverse impacts from military aviation noise. Results highlight the need for improved noise metrics, modeling and operational adjustments to mitigate the most disruptive effects.
SERONEY, G. C.; Magak, N. A. G.; Mchunu, G. G.
Show abstract
Introduction Access to safe water, sanitation, and hygiene (WASH) in schools is critical for child health, learning, and gender equity. In Kenya, the Kenya School Health Policy and the Basic Education Act outline standards for school WASH; however, implementation remains uneven due to inadequate infrastructure, weak inter-sectoral coordination, and limited financing. This study aimed to identify priority components for strengthening school WASH implementation and generate policy-relevant recommendations based on expert consensus in Uasin Gishu County, Kenya. Methods and Results A Delphi technique consisting of two iterative rounds was used to reach expert consensus. In Round 1, 20 purposively selected experts including head teachers, county education officials, public health officers, water and public works officers, and NGO representatives participated in key informant interviews. Emergent themes informed development of a structured Round 2 questionnaire administered through CommCare online app. Quantitative data were analyzed using descriptive statistics (means, standard deviations, percentage agreement), while qualitative responses underwent thematic coding using NVivo 12. Experts reached strong consensus on essential components required for strengthening school WASH implementation. Key priorities included clear governance structures, designated budget lines, inclusive infrastructure, menstrual hygiene management (MHM), curriculum integration, sustained capacity building, and systematic monitoring. Multi-sectoral collaboration and recognition of best-performing schools were also emphasized as important motivators for compliance and sustainability. Equity considerations particularly the need for disability-friendly facilities and school-community outreach were highlighted as critical. Agreement levels ranged from 74% to 100%, with most items scoring mean values between 4.5 and 4.8 on a 5-point Likert scale, indicating strong consensus among experts. Conclusion strengthening implementation of school WASH in Kenya requires coordinated governance, predictable funding, reliable water systems, inclusive sanitation, strengthened MHM, and consistent monitoring beyond infrastructure investment alone. Integrating these expert-validated priorities within existing national policies offers a practical pathway to improving learner health, reducing absenteeism especially among girls and promoting equitable educational outcomes.
Yu, J.; Tillema, S.; Akel, M.; Aron, A.; Espinosa, E.; Fisher, S. A.; Branche, T. N.; Mithal, L. B.; Hartmann, E. M.
Show abstract
Benzalkonium chloride (BAC) is widely used as a disinfectant in cleaning products and is frequently detected in indoor dust. In this study, we assessed dust samples, along with information on cleaning product use, from 24 pregnant participants. Dust samples were analyzed for BAC concentration and microbial tolerance. Different chain lengths of BAC (C12, C14, and C16) were quantified using LC-MS/MS, and bacterial isolates were tested for BAC tolerance using minimum inhibitory concentration (MIC) assays. BAC was ubiquitously detected, with C12 and C14 being dominant. Higher BAC concentrations were associated with reported disinfectant use and increased microbial tolerance. These findings suggest that indoor antimicrobial use may promote microbial resistance, highlighting potential exposure risks in indoor environments and the need for further investigation into health and ecological impacts.
Salim, A.; Allen, M.; Mariki, K.; Pallangyo, T.; Maina, R.; Mzee, F.; Minja, M.; Msovela, K.; Liana, J.
Show abstract
In the context of global health, the ability of frontline primary health providers to identify potential Drug-Drug Interactions (DDIs) is a critical component of patient safety. This is particularly true in settings like Tanzania, where drug dispensers often serve as the primary point of contact for healthcare. In this study, we establish a baseline for drug decision-making capabilities across multiple cadres of healthcare providers in Kibaha, Tanzania. We specifically distinguish between the ability to recognize safe drug combinations versus harmful ones. The findings reveal a critical asymmetry in provider performance: while professional training improves the recognition of safe combinations, it provides no advantage over lay intuition (and in some cases, a significant disadvantage) in detecting potentially harmful interactions.
Strand, P. S.; Trang, J. C.
Show abstract
Female genital cutting (FGC) is identified within global health and human rights discourse as aligned with gender inequality and female disempowerment. The persistence of FGC in high-prevalence societies is assumed to reflect womens limited influence over decisions concerning their daughters. Yet anthropological research has questioned whether this interpretation adequately reflects how FGC is organized within practicing communities. Across two studies with 176,728 participants from 15 African and Asian countries, we examine whether mothers attitudes toward FGC predict daughters circumcision status and whether this relationship varies with regional FGC prevalence. Multilevel logistic regression models show that maternal attitudes strongly predict daughter circumcision status across both datasets. Contrary to expectations derived from disempowerment frameworks, the association between maternal attitudes and daughter outcomes is not weaker in high-prevalence contexts, it is stronger. These findings suggest that interpretations of FGC as reflecting female disempowerment may mischaracterize the social dynamics of societies in which FGC is common. Policy implications of the findings are discussed.
Masegese, T.; MUNG'ONG'O, G. S.; Kamala, B.; Anaeli, A.; Bago, M.; Mtoro, M. J.
Show abstract
Background: HIV/AIDS remains a major public health challenge in Tanzania, where viral load suppression among adults on ART stands at 78% and HVL testing uptake among eligible patients is approximately 22%. Since the introduction of the National HVL Testing Guideline in 2015, little has been done to systematically evaluate its implementation. Objective: To evaluate adherence to the National HVL Testing Guideline across CTC clinics in Dar es Salaam Region, covering ART monitoring, documentation, turnaround time, and factors affecting implementation. Methods: A cross-sectional study was conducted in 2021 across 15 public health facilities with CTC clinics in all five Dar es Salaam districts. A total of 330 PLHIV on ART for more than six months were selected through systematic random sampling with proportional to size allocation, and 45 healthcare providers through convenient sampling. Data were collected via abstraction forms and self-administered questionnaires, and analysed using SPSS Version 23 with descriptive statistics, bivariate analysis, and binary logistic regression. Results: Only 25.1% of patients had their first HVL sample taken at six months as per guideline, with 68.8% delayed beyond six months. Second and third samples were similarly delayed. MoHCDGEC sample tracking forms were absent in 96.7% of facilities and incomplete in 99.1%, and no facility captured specimen acceptance or rejection as site feedback. Turnaround time exceeded the 14-day guideline threshold in 64.5%, 66.7%, and 69.4% of first, second, and third results respectively. Patient negligence (AOR=9.84; 95% CI: 1.83-52.77) and storage (AOR=5.72; 95% CI: 0.94-35.0) were independently associated with guideline adherence. Conclusion: Adherence to the National HVL Testing Guideline in Dar es Salaam is suboptimal across testing timelines, documentation, and turnaround time, with patient negligence and storage capacity as significant determinants. Targeted interventions are needed to strengthen patient education, improve storage infrastructure, enhance documentation systems, and support providers in adhering to guideline-specified timelines.
Henderson, D.; Lignier, B.; Moxham, B.; Plaisant, O.; OSCEs study group, U. P. C.; Buffel du Vaure, C.; Faye, A.; Bouzid, D.; Lemogne, C.; Guedon, A.
Show abstract
ABSTRACT Objective Structured Clinical Examinations (OSCEs) are widely used to assess medical students clinical skills, including non-technical abilities such as communication and empathy. However, the potential influence of individual psychological traits, such as personality dimensions, empathy, and stress-related mindset on OSCE performance remains understudied. This study investigated associations between personality traits, empathy levels, stress mindsets, and performance in OSCEs among medical students. An online questionnaire (including the Big Five Personality Traits Inventory 2, the Jefferson Scale of Physician Empathy (Medical Student version), the Growth Mindset Scale, the Stress Mindset Measure) was provided to all fifth-year medical students enrolled at the Universite Paris Cite for six weeks before undertaking graduation summative OSCEs. Their scores were correlated with OSCE performance using Spearmans correlation and linear regression analyses. A total of 99 questionnaires were included and analysed. None of the psychometric tests we assessed showed a significant correlation with OSCE scores. The strongest predictors of success in OSCEs were higher scores in written examinations, previous OSCE performance, and being female. In non-interactive OSCE stations, conscientiousness was the only significant predictor, with a positive association (p=0.001). Neuroticism was positively associated with performance improvement between OSCE sessions (p=0.042). Personality traits, self-reported empathy, and stress-related mindsets do not predict success in OSCEs as isolated traits. Further research is needed to determine whether it holds true for all kinds of OSCEs. Multidimensional psychometric assessment may be relevant when investigating performance outcomes in OSCEs.
Trivedi, S.; Simons, N. W.; Tyagi, A.; Ramaswamy, A.; Nadkarni, G. N.; Charney, A. W.
Show abstract
Background: Large language models (LLMs) are increasingly used in mental health contexts, yet their detection of suicidal ideation is inconsistent, raising patient safety concerns. Objective: To evaluate whether an independent safety monitoring system improves detection of suicide risk compared with native LLM safeguards. Methods: We conducted a cross-sectional evaluation using 224 paired suicide-related clinical vignettes presented in a single-turn format under two conditions (with and without structured clinical information). Native LLM safeguard responses were compared with an independent supervisory safety architecture with asynchronous monitoring. The primary outcome was detection of suicide risk requiring intervention. Results: The supervisory system detected suicide risk in 205 of 224 evaluations (91.5%) versus 41 of 224 (18.3%) for native LLM safeguards. Among 168 discordant evaluations, 166 favored the supervisory system and 2 favored the LLM (matched odds ratio {approx}83.0). Both systems detected risk in 39 evaluations, and neither in 17. Detection was highest in scenarios with explicit suicidal ideation and lower in more ambiguous presentations. Conclusions: Native LLM safeguards frequently failed to detect suicide risk in this structured evaluation. An independent monitoring approach substantially improved detection, supporting the role of external safety systems in high-risk mental health applications of LLMs.
Monson, E. T.; Shabalin, A. A.; Diblasi, E.; Staley, M. J.; Kaufman, E. A.; Docherty, A. R.; Bakian, A. V.; Coon, H.; Keeshin, B. R.
Show abstract
Importance: Suicide is a leading cause of death in the United States with risk strongly influenced by Interpersonal trauma, contributing to treatment resistance and clinical complexity. Objective: To assess clinical and genetic factors in individuals who died from suicide, with and without interpersonal trauma exposure. Design: Individuals who died from suicide with and without trauma were compared in a retrospective case-case design. Prevalence of 19 broad clinical categories was assessed between groups. Results directed selection of 42 clinical subcategories, and 40 polygenic scores (PGS) for further assessment. Multivariable logistic regression models, adjusted for critical covariates and multiple tests, were formulated. Models were also stratified by age group (<26yo and >=26yo), sex, and age/sex. Setting: A population-based evaluation of comorbidity and polygenic scoring in two suicide death subgroups. Participants: A total of 8 738 Utah Suicide Mortality Research Study individuals (23.9% female, average age = 42.6 yo) who died from suicide were evaluated, divided into trauma (N = 1 091) and non-trauma exposed (N = 7 647) individuals. A subset of unrelated European genotyped individuals was also assessed in PGS analyses (Trauma N = 491; Non-trauma N = 3 233). Exposures: Trauma is here defined as interpersonal trauma exposure, including abuse, assault, and neglect from International Classification of Disease coding. Main Outcomes and Measures: Prevalence of comorbid clinical sub/categories and PGS enrichment in trauma versus non-trauma exposed suicide deaths. Results: Overall, trauma-exposed individuals died from suicide earlier (mean age of 38.1 yo versus 43.3 yo; P <0.0001) and were disproportionately female (38% versus 21%, OR = 3.3, CI = 2.9-3.8). Prevalence of asphyxiation and overdose methods, prior suicidality, psychiatric diagnoses, and substance use (OR range = 1.3-3.7) were elevated in trauma exposed individuals who died from suicide. Genetic PGS were also elevated in trauma-exposed individuals who died from suicide for depression, bipolar disorder, cannabis use, PTSD, insomnia, and schizophrenia (OR range = 1.1-1.4) with ADHD and opioid use showing uniquely elevated PGS in trauma exposed males (OR range = 1.2-1.4). Conclusions and Relevance: Results demonstrated multiple convergent lines of age- and sex-specific evidence differentiating trauma-exposed from non-trauma exposed suicide death. Such findings suggest unique biological backgrounds and may refine identification and treatment of this high-risk group.
Umar, M.; Hussain, F.; Khizar, B.; Khan, I.; Khan, F.; Cotic, M.; Chan, L.; Hussain, A.; Ali, M. N.; Gill, S. A.; Mustafa, A. B.; Dogar, I. A.; Nizami, A. T.; Haq, M. M. u.; Mufti, K.; Ansari, M. A.; Hussain, M. I.; Choudhary, S. T.; Maqsood, N.; Rasool, G.; Ali, H.; Ilyas, M.; Tariq, M.; Shafiq, S.; Khan, A. A.; Rashid, S.; Ahmad, H.; Bettani, K. U.; Khan, M. K.; Choudhary, A. R.; Mehdi, M.; Shakoor, A.; Mehmood, N.; Mufti, A. A.; Bhatia, M. R.; Ali, M.; Khan, M. A.; Alam, N.; Naqvi, S. Q.-i.-H.; Mughal, N.; Ilyas, N.; Channar, P.; Ijaz, P.; Din, A.; Agha, H.; Channa, S.; Ambreen, S.; Rehman,
Show abstract
BackgroundMajor depressive disorder (MDD), a leading cause of disability worldwide, exhibits substantial heterogeneity in treatment outcomes. Patients who do not respond to standard antidepressant therapy account for the majority of MDDs disease burden. Risk factors have been implicated in treatment response, including genes impacting on how antidepressants are metabolised. Yet, despite its clinical importance, risk factors for treatment-resistant depression (TRD) remain unexplored in low- and middle-income countries (LMIC). We used data from the DIVERGE study on MDD to investigate the risk factors of TRD in Pakistan. MethodsDIVERGE is a genetic epidemiological study that recruited adult MDD patients ([≥]18 years) between Sep 27,2021 to Jun 30, 2025, from psychiatric care facilities across Pakistan. Detailed phenotypic information was collected by trained interviewers and blood samples taken. Infinium Global Diversity Array with Enhanced PGx-8 from Illumina was used for genotyping followed by DRAGEN calling to infer metaboliser phenotypes for Cytochrome P450 (CYP) enzyme genes. We defined TRD as minimal to no improvement after [≥]12 weeks of adherent antidepressant therapy. We conducted multi-level logistic regression to test the association of demographic, clinical and pharmacogenetic variables with TRD. FindingsAmong 3,677 eligible patients, polypharmacy was rampant; 86% were prescribed another psychotropic drug along with an antidepressant. Psychological therapies were uncommon (6%) while 49% of patients had previously visited to a religious leader/faith healer in relation to their mental health problems. TRD was experienced by 34% (95%CI: 32-36%) patients. The TRD group was characterised by more psychotic symptoms and suicidal behaviour (OR=1.39, 95%CI=1.04-1.84, p=0.02; OR=1.03, 95%CI=1.01-1.05, p=0.005). Social support (OR=0.55, 95%CI=0.44-0.69, p=1.4x10-7) and parents being first cousins (OR=0.81, 95%CI=0.69-0.96, p=0.01) were associated with lower odds of TRD. In 1,085 patients with CYP enzyme data, poor (OR=1.85, 95%CI=1.11-3.07, p=0.01) and ultra-rapid (OR=3.11, 95%CI=1.59-6.12, p=0.0009) metabolizers for CYP2C19 had increased risk of TRD compared with normal metabolisers. InterpretationThere was an excessive use of polypharmacy in the treatment of depression while psychological therapies were uncommon highlighting the need for more evidence-based practice. This first large study of MDD from Pakistan uncovered the importance of culture-specific forms of social support in preventing TRD, highlighting opportunities for interventions in low-income settings. Pharmacogenetic markers can be leveraged to predict TRD.
Mutibwa, S.; Wandiembe, S.; Mbonye, K.; Nsimbe, D.
Show abstract
Background: Preterm births contribute to approximately 35% of neonatal deaths globally, with an estimated 13.4 million infants born prematurely each year. Despite this substantial burden, limited evidence exists on time to discharge and its determinants among preterm neonates admitted to Neonatal Intensive Care Units (NICUs), particularly in rural Ugandan settings. This study aimed to investigate time to discharge and associated factors among preterm neonates admitted to Kiwoko Hospital in Nakaseke District, Uganda. Methods: A retrospective cohort study was conducted using secondary data from Kiwoko Hospital on preterm neonates admitted to the Neonatal Intensive Care Unit (NICU) between 2020 and 2021 (n = 847). The cumulative incidence function was used to estimate the probability of discharge within 28 days of admission, accounting for competing events. A Fine and Gray sub-distribution hazard regression model was fitted to identify factors associated with time to discharge. Results: Of the 847 preterm admissions, 70.1% were discharged alive within 28 days. The median time to discharge was 14 days. The cumulative incidence of discharge by 28 days was 68%, accounting for competing events. During follow-up, 165 neonates did not complete the 28-day period, including 88 deaths. Factors significantly associated with time to discharge included place of delivery (SHR: 0.62; 95% CI: 0.53-0.73; p<0.001), maternal residence in other districts (SHR: 0.69; 95% CI: 0.48-0.99; p=0.044), extreme preterm (SHR: 0.05; 95% CI: 0.03-0.09; p<0.001), very preterm (SHR: 0.18; 95% CI: 0.14-0.25; p<0.001), moderate preterm (SHR: 0.59; 95% CI: 0.46-0.76; p<0.001), triplet births (SHR: 0.40; 95% CI: 0.23-0.68; p=0.001), 2-4 ANC visits (SHR: 0.70; 95% CI: 0.56-0.87; p=0.002), <=1 ANC visit (SHR: 0.64; 95% CI: 0.49-0.85; p=0.002), respiratory distress syndrome (SHR: 0.64; 95% CI: 0.48-0.74; p<0.001), and birth trauma (SHR: 2.62; 95% CI: 1.60-4.29; p<0.001). Conclusions: Respiratory distress syndrome, fewer antenatal care visits, out-of-district residence, and higher degrees of prematurity were associated with prolonged time to discharge among preterm neonates. Strengthening antenatal care utilization and improving access to quality neonatal care in underserved areas may enhance discharge outcomes.
Weill, O.; Lucas, N.; Bailey, B.; Marquis, C.; Gravel, J.
Show abstract
Objectives: Acute gastroenteritis is a leading cause of pediatric emergency department (ED) visits. While ondansetron reduces vomiting, intravenous rehydration, and hospital admissions, its efficacy when initiated at triage remains unclear. We aimed to evaluate whether triage nurse-initiated administration of ondansetron in children with suspected gastroenteritis reduces the proportion of patients requiring observation following initial physician assessment. Methods: We conducted a randomized, double-blind, placebo-controlled trial in a tertiary pediatric ED in Canada. Children aged 6 months to 17 years presenting with morae than 3 episodes of vomiting in the preceding 24 hours (including 1 within 2 hours of arrival), were eligible. At triage, we randomized participants to receive liquid ondansetron or a color- and taste-matched placebo. The primary outcome was the proportion of patients requiring observation after the first physician evaluation. Secondary outcomes included post-intervention vomiting, ED length of stay, patient comfort, and 48-hour return visits. The trial was registered at ClinicalTrials.gov (NCT03052361). Results: Recruitment was stopped prematurely due to the COVID-19 pandemic. Ninety-one participants were randomized to ondansetron (n= 44) or placebo (n= 47). Overall, 40 patients (45%) were discharged immediately after the initial physician assessment, with no difference between the ondansetron and placebo groups (44% vs. 45%; absolute difference -1%, 95% CI: -20% to 19%). No significant differences were observed in all secondary outcomes. Conclusion: In this trial, triage nurse-initiated ondansetron administration did not reduce the need for ED observation in children with presumed gastroenteritis. While being underpowered, this study could inform researchers planning larger clinical trials.
Aravamuthan, B. R.; Bailes, A. F.; Baird, M.; Bjornson, K.; Bowen, I.; Bowman, A.; Boyer, E.; Gelineau-Morel, R.; Glader, L.; Gross, P.; Hall, S.; Hurvitz, E.; Kruer, M. C.; Larrew, T.; Marupudi, N.; McPhee, P.; Nichols, S.; Noritz, G.; Oleszek, J.; Ramsey, J.; Raskin, J.; Riordan, H.; Rocque, B.; Shah, M.; Shore, B.; Shrader, M. W.; Spence, D.; Stevenson, C.; Thomas, S. P.; Trost, J.; Wisniewski, S.
Show abstract
Objective Cerebral palsy (CP) affects approximately 1 million Americans and 18 million individuals worldwide, yet contemporary US epidemiologic data remains limited. We aimed to use Cerebral Palsy Research Network (CPRN) clinical registry to describe demographics and clinical characteristics of individuals with CP across the US and determine associations with gross motor function and genetic etiology. Methods Registry subjects were included if they had clinician-confirmed CP and prospectively entered data for Gross Motor Function Classification System (GMFCS) Level, gestational age, genetic etiology, CP distribution, and tone/movement types. Logistic regression was used to determine which of these variables plus race, sex, ethnicity, and age were associated with GMFCS level and genetic etiology. Results A total of 9,756 children and adults with CP from 22 CPRN sites met inclusion criteria. Participants were predominantly White (73.0%), male (57.3%), non-Hispanic (87.8%), and younger than 18 years (73.7%). Most were classified as GMFCS levels I-III (55.6%), born preterm (52.8%), had spasticity (83.8%), and had quadriplegia (41.9%); 12.2% were identified as having a genetic etiology. Tone/movement types, CP distribution, and gestational age were significantly associated with both GMFCS level and genetic etiology (p<0.001). Compared to White individuals, Black individuals were more likely to have greater gross motor impairment (p<0.001). Conclusion In this large US cohort, clinical and demographic factors, including race, were associated with gross motor function and genetic etiology in CP. These findings highlight persistent disparities and demonstrate the value of a national clinical registry for informing prognostication, quality improvement efforts, and targeted genetic testing strategies.
Bannett, Y.; Pillai, M.; Huang, T.; Luo, I.; Gunturkun, F.; Hernandez-Boussard, T.
Show abstract
ImportanceGuideline-concordant care for young children with attention-deficit/hyperactivity disorder (ADHD) includes recommending parent training in behavior management (PTBM) as first-line treatment. However, assessing guideline adherence through manual chart review is time-consuming and costly, limiting scalable and timely quality-of-care measurement. ObjectiveTo evaluate the accuracy and explainability of large language models (LLMs) in identifying PTBM recommendations in pediatric electronic health record (EHR) notes as a scalable alternative to manual chart review. Design, Setting, and ParticipantsThis retrospective cohort study was conducted in a community-based pediatric healthcare network in California consisting of 27 primary care clinics. The study cohort included children aged 4-6 years with [≥] 2 primary care visits between 2020-2024 and ICD-10 diagnoses of ADHD or ADHD symptoms (n=542 patients). Clinical notes from the first ADHD-related visit were included. A stratified subset of 122 notes, including all cases with model disagreement, was manually annotated to assess model performance in identifying PTBM recommendations and rank model explanations. ExposuresAssessment and plan sections of clinical notes were analyzed using three generative large language models (Claude-3.5, GPT-4o, and LLaMA-3.3-70B) to identify the presence of PTBM recommendations and generate explanatory rationales and documentation evidence. Main Outcomes and MeasuresModel performance in identifying PTBM recommendations (measured by sensitivity, positive predictive value (PPV), and F1-score) and qualitative explainability ratings of model-generated rationales (based on the QUEST framework). ResultsAll three models demonstrated high performance compared to expert chart review. Claude-3.5 showed balanced performance (sensitivity=0.89, PPV=0.95, and F1-score=0.92) and ranked highest in explainability. LLaMA3.3-70B achieved sensitivity=0.91, PPV=0.89, and F1-score=0.90, ranking second for explainability. GPT-4o had the highest PPV [0.97] but lowest sensitivity [0.82], with an F1-score of 0.89 and the lowest explainability ranking. Based on classifications from the best-performing model, Claude-3.5, 26.4% (143/542) of patients had documented PTBM recommendations at their first ADHD-related visit. Conclusions and RelevanceLLMs can accurately extract guideline-concordant clinician recommendations for non-pharmacological ADHD treatment from unstructured clinical notes while providing clear explanations and supporting evidence. Evaluating model explainability as part of LLM implementation for medical chart review tasks can promote transparent and scalable solutions for quality-of-care measurement.
Camara, S. M. A.; de Souza Barbosa, J. F.; Hipp, S.; Fernandes Macedo, S. G. G.; Sentell, T.; Bassani, D. G.; Domingues, M. R.; Pirkle, C. M.
Show abstract
BackgroundProspective studies of pregnant adolescents are essencial to effectively address this global health priority. They help answer vital questions about their health, but such studies are uncommon due to the difficulty in retaining adolescents. This paper describes the successes and challenges of the research strategies used to ensure sufficient recruitment and retention of pregnant adolescents in a longitudinal study about adolescent childbearing in an under-resourced setting. MethodsThe Adolescence and Motherhood Research project was conducted in a rural region of Northeast Brazil in 2017-2019 and assessed 50 primigravids between 13-18 years (adolescents) and 50 primigravids between 23-28 years (young adults) during the first 16 weeks of pregnancy with two follow-ups (third trimester of pregnancy, and 4-6 weeks postpartum). Recruitment strategies involved engagement of health sector and community, as well as referrals from health care professionals and dissemination of the project in different locations. Retention strategies included maintaining contact with the participants between assessments and providing transportation for them to attend the follow-up procedures. ResultsRecruitment took 10 months to complete. A total of 78% of the participants were recruited from the primary health care units, mainly after referral from a health care provider. Retention reached 95% of the sample throughout the study (90%: adolescents; 98%: adults). ConclusionA combination of approaches is necessary to successfully recruit and retain youth in longitudinal studies and engaging local stakeholders may help to increase community-perceived legitimacy of the research. Working closely with front-line staff is essential when conducting research in rural low-income communities.
Zhang, Q.; Tang, Q.; Vu, T.; Pandit, K.; Cui, Y.; Yan, F.; Wang, N.; Li, J.; Yao, A.; Menozzi, L.; Fung, K.-M.; Yu, Z.; Parrack, P.; Ali, W.; Liu, R.; Wang, C.; Liu, J.; Hostetler, C. A.; Milam, A. N.; Nave, B.; Squires, R. A.; Battula, N. R.; Pan, C.; Martins, P. N.; Yao, J.
Show abstract
End-stage liver disease (ESLD) is one of the leading causes of death worldwide. Currently, the only curative option for patients with ESLD is liver transplantation. However, the demand for donor livers far exceeds the available supply, partly because many potentially viable livers are discarded following biopsy evaluation. While biopsy is the gold standard for assessing liver histological features related to graft quality and transplant suitability, it often leads to high discard rates due to its susceptibility to sampling errors and limited spatial coverage. Besides, biopsy is invasive, time-consuming, and unavailable in clinical facilities with limited resources. Here, we present an AI-assisted photoacoustic/ultrasound (PA/US) imaging framework for quantitative assessment of human donor liver graft quality and transplant suitablity at the whole-organ scale. With multimodal volumetric PA/US images as the input, our deep-learning (DL) model accurately predicted the risk level of fibrosis and steatosis, which indicate the graft quality and transplant suitability, when comparing with true pathological scores. DL also identified the imaging modes (PAI wavelength and B-mode USI) that correlated the most with prediction accuracy, without relying on ill-posed spectral unmixing. Our method was evaluated in six discarded human donor livers comprising sixty spatially matched regions of interest. Our study will pave the way for a new standard of care in organ graft quality and transplant suitability that is fast, noninvasive, and spatially thorough to prevent unnecessary organ discards in liver transplantation.
Xu, R.; Dou, H.; Zhang, M.; Liu, Z.
Show abstract
Background: To investigate the safety and efficacy of CTguided lung nodule localization needles for the preoperative localization of small pulmonary nodules. Methods: A retrospective study was conducted on 102 patients with a total of 113 small pulmonary nodules who underwent preoperative localization at Jinan Fourth People's Hospital from January 2024 to December 2025. Nodule diameter and depth, localization time, the number of pleural punctures, the localization success rate, and postoperative complications (hook dislodgement, hemorrhage, and pneumothorax) were recorded. All patients underwent video assisted thoracoscopic surgery (VATS) after localization. Results: The mean nodule diameter was 0.97{+/-}0.36 cm, the mean depth was 1.26{+/-}0.48 cm, and the mean localization time was 9.8{+/-}3.65 minutes. The hook dislodgement rate was 0.98% (1/102), the intrapulmonary hemorrhage rate was 14.71% (15/102), and the pneumothorax rate was 16.67% (17/102). All pulmonary nodules were successfully resected by VATS at 73.82{+/-}13.83 minutes after localization, and no severe complications occurred. Conclusions: The use of a CTguided lung nodule localization needle for the preoperative localization of small pulmonary nodules decreases the time needed for intraoperative nodule detection and operation time. This strategy is a simple, safe, and accurate preoperative localization method that is worthy of increased clinical use.